Goto

Collaborating Authors

 anytime deterministic guarantee


Online POMDP Planning with Anytime Deterministic Guarantees

Neural Information Processing Systems

Autonomous agents operating in real-world scenarios frequently encounter uncertainty and make decisions based on incomplete information. Planning under uncertainty can be mathematically formalized using partially observable Markov decision processes (POMDPs). However, finding an optimal plan for POMDPs can be computationally expensive and is feasible only for small tasks. In recent years, approximate algorithms, such as tree search and sample-based methodologies, have emerged as state-of-the-art POMDP solvers for larger problems. Despite their effectiveness, these algorithms offer only probabilistic and often asymptotic guarantees toward the optimal solution due to their dependence on sampling.


Online POMDP Planning with Anytime Deterministic Guarantees - Supplementary Anonymous Author(s) Affiliation Address email

Neural Information Processing Systems

W e restate some of the definitions from the paper for convenience. Moreover, depending on the algorithm implementation, the number of iterations can be finite (e.g. by After the allowable time steps ended, the simulation was reset to its initial state. The main goal is to tag as many opponents as possible within a given time frame. The states reflect the agent's current position and the opponents' positions. The Baby POMDP is a classic problem that represents the scenario of a baby and a caregiver. The states in this problem represent the baby's needs, which could be hunger, The observations are binary, either the baby is crying or not.


Online POMDP Planning with Anytime Deterministic Guarantees

Neural Information Processing Systems

Autonomous agents operating in real-world scenarios frequently encounter uncertainty and make decisions based on incomplete information. Planning under uncertainty can be mathematically formalized using partially observable Markov decision processes (POMDPs). However, finding an optimal plan for POMDPs can be computationally expensive and is feasible only for small tasks. In recent years, approximate algorithms, such as tree search and sample-based methodologies, have emerged as state-of-the-art POMDP solvers for larger problems. Despite their effectiveness, these algorithms offer only probabilistic and often asymptotic guarantees toward the optimal solution due to their dependence on sampling.


Online POMDP Planning with Anytime Deterministic Guarantees

Barenboim, Moran, Indelman, Vadim

arXiv.org Artificial Intelligence

Autonomous agents operating in real-world scenarios frequently encounter uncertainty and make decisions based on incomplete information. Planning under uncertainty can be mathematically formalized using partially observable Markov decision processes (POMDPs). However, finding an optimal plan for POMDPs can be computationally expensive and is feasible only for small tasks. In recent years, approximate algorithms, such as tree search and sample-based methodologies, have emerged as state-of-the-art POMDP solvers for larger problems. Despite their effectiveness, these algorithms offer only probabilistic and often asymptotic guarantees toward the optimal solution due to their dependence on sampling. To address these limitations, we derive a deterministic relationship between a simplified solution that is easier to obtain and the theoretically optimal one. First, we derive bounds for selecting a subset of the observations to branch from while computing a complete belief at each posterior node. Then, since a complete belief update may be computationally demanding, we extend the bounds to support reduction of both the state and the observation spaces. We demonstrate how our guarantees can be integrated with existing state-of-the-art solvers that sample a subset of states and observations. As a result, the returned solution holds deterministic bounds relative to the optimal policy. Lastly, we substantiate our findings with supporting experimental results.